No More Pesky Learning Rates: Supplementary Material
نویسندگان
چکیده
If we do gradient descent with η * (t), then almost surely, the algorithm converges (for the quadratic model). To prove that, we follow classical techniques based on Lyapunov stability theory (Bucy, 1965). Notice that the expected loss follows E J θ (t+1) | θ (t) = 1 2 h · E (1 − η * h)(θ (t) − θ *) + η * hσξ 2 + σ 2 = 1 2 h (1 − η * h) 2 (θ (t) − θ *) 2 + (η *) 2 h 2 σ 2 + σ 2 = 1 2 h σ 2 (θ (t) − θ *) 2 + σ 2 (θ (t) − θ *) 2 + σ 2 ≤ J θ (t) Thus J(θ (t)) is a positive super-martingale, indicating that almost surely J(θ (t)) → J ∞. We are to prove that almost surely J ∞ = J(θ *) = 1 2 hσ 2. Observe that J(θ (t)) − E[J(θ (t+1)) | θ (t) ] = 1 2 hη * (t) , E[J(θ (t))] − E[J(θ (t+1)) | θ (t) ] = 1 2 hE[η * (t)] Since E[J(θ (t))] is bounded below by 0, the telescoping sum gives us E[η * (t)] → 0, which in turn implies that in probability η * (t) → 0. We can rewrite this as η * (t) = J(θ t) − 1 2 hσ 2 J(θ t) → 0 By uniqueness of the limit, almost surely,
منابع مشابه
No more pesky learning rates
The performance of stochastic gradient descent (SGD) depends critically on how learning rates are tuned and decreased over time. We propose a method to automatically adjust multiple learning rates so as to minimize the expected error at any one time. The method relies on local gradient variations across samples. In our approach, learning rates can increase as well as decrease, making it suitabl...
متن کاملInverse Reinforcement Learning with Simultaneous Estimation of Rewards and Dynamics - Supplementary Material
This document contains supplementary material to the paper Inverse Reinforcement Learning with Simultaneous Estimation of Rewards and Dynamics with more detailed derivations, additional proofs to lemmata and theorems as well as larger illustrations and plots of the evaluation task. 1 Partial Derivative of the Policy
متن کاملSupplementary Materials for Learning to Recognize Discontiguous Entities
This is the supplementary material for “Learning to Recognize Discontiguous Entities” [Muis and Lu, 2016]. This material gives more details in the experiment setup, the ambiguity of each model, and compare the models from theoretical point of view.
متن کاملPart-based Visual Tracking with Online Latent Structural Learning: Supplementary Material
In this supplementary material, we provide more related work, more details of structured output tracking and online latent structural learning for visual tracking, and more experimental results (both qualitative and quantitative). The experimental results includes demonstration videos, more quantitative results of our tracking algorithm with and without parts in Tab. 2 and Fig. 1, more quantita...
متن کاملIdentification of the Quantitative Dimensions of Supplementary Private Tutoring among High School Students
Identification of the Quantitative Dimensions of Supplementary Private Tutoring among High School Students Y. Shiraavand* A. Ma’dandaar Aaraani, Ph.D.** I. Nikpey, Ph.D.*** To identify the quantitative dimensions of supplementary private tutoring, a cluster sample of 800 high school students was given a questionnaire on the subject. The collected data reveal that moti...
متن کامل